# CLIP architecture
Eva Giant Patch14 Clip 224.laion400m
MIT
The EVA CLIP model is a vision-language model based on OpenCLIP and the timm framework, supporting zero-shot image classification tasks.
Text-to-Image
E
timm
124
0
Eva02 Enormous Patch14 Clip 224.laion2b Plus
MIT
EVA-CLIP is a large-scale vision-language model based on the CLIP architecture, supporting tasks such as zero-shot image classification.
Text-to-Image
E
timm
54
0
Eva02 Enormous Patch14 Clip 224.laion2b
MIT
EVA-CLIP is a vision-language model based on the CLIP architecture, supporting zero-shot image classification tasks.
Text-to-Image
E
timm
38
0
Eva02 Base Patch16 Clip 224.merged2b
MIT
The EVA CLIP model is a vision-language model built on the OpenCLIP and timm frameworks, supporting tasks like zero-shot image classification.
Text-to-Image
E
timm
3,029
0
Resnet50 Clip.openai
MIT
Zero-shot image classification model based on ResNet50 architecture and CLIP technology
Image Classification
R
timm
11.91k
0
Mobileclip S0
Other
MobileCLIP S0 is an ONNX adaptation of Apple's ml-mobileclip project, a zero-shot image classification model optimized for mobile devices.
Text-to-Image
Transformers

M
Xenova
295
1
Vit Betwixt Patch32 Clip 224.tinyclip Laion400m
MIT
A small CLIP model based on ViT architecture, suitable for zero-shot image classification tasks, trained on the LAION-400M dataset.
Image Classification
V
timm
113
1
Vit Medium Patch32 Clip 224.tinyclip Laion400m
MIT
A vision-language model based on the OpenCLIP library, supporting zero-shot image classification tasks.
Image Classification
V
timm
110
0
Vit Medium Patch16 Clip 224.tinyclip Yfcc15m
MIT
CLIP model based on ViT architecture for zero-shot image classification tasks
Image Classification
V
timm
144
0
CLIP ViT L 14 CommonPool.XL S13b B90k
MIT
A vision-language pretrained model based on the CLIP architecture, supporting zero-shot image classification and cross-modal retrieval tasks
Text-to-Image
C
laion
4,255
2
Featured Recommended AI Models